Online Hyperparameter Optimization for Class-Incremental Learning

نویسندگان

چکیده

Class-incremental learning (CIL) aims to train a classification model while the number of classes increases phase-by-phase. An inherent challenge CIL is stability-plasticity tradeoff, i.e., models should keep stable retain old knowledge and plastic absorb new knowledge. However, none existing can achieve optimal tradeoff in different data-receiving settings—where typically training-from-half (TFH) setting needs more stability, but training-from-scratch (TFS) plasticity. To this end, we design an online method that adaptively optimize without knowing as priori. Specifically, first introduce key hyperparameters influence e.g., distillation (KD) loss weights, rates, classifier types. Then, formulate hyperparameter optimization process Markov Decision Process (MDP) problem propose specific algorithm solve it. We apply local estimated rewards classic bandit Exp3 address issues when applying MDP methods protocol. Our consistently improves top-performing both TFH TFS settings, boosting average accuracy by 2.2 percentage points on ImageNet-Full, compared state-of-the-art. Code provided at https://class-il.mpi-inf.mpg.de/online/

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Incremental Class Dictionary Learning and Optimization

We have previously shown how the discovery of classes from objects can be automated, and how the resulting class organization can be eeciently optimized in the case where the optimum is a single inheritance class hierarchy. This paper extends our previous work by showing how an optimal class dictionary can be learned incrementally. The ability to expand a class organization incrementally as new...

متن کامل

Bayesian Hyperparameter Optimization for Ensemble Learning

In this paper, we bridge the gap between hyperparameter optimization and ensemble learning by performing Bayesian optimization of an ensemble with regards to its hyperparameters. Our method consists in building a fixed-size ensemble, optimizing the configuration of one classifier of the ensemble at each iteration of the hyperparameter optimization algorithm, taking into consideration the intera...

متن کامل

Initializing Bayesian Hyperparameter Optimization via Meta-Learning

Model selection and hyperparameter optimization is crucial in applying machine learning to a novel dataset. Recently, a subcommunity of machine learning has focused on solving this problem with Sequential Model-based Bayesian Optimization (SMBO), demonstrating substantial successes in many applications. However, for computationally expensive algorithms the overhead of hyperparameter optimizatio...

متن کامل

Learning to Warm-Start Bayesian Hyperparameter Optimization

Hyperparameter optimization undergoes extensive evaluations of validation errors in order to find the best configuration of hyperparameters. Bayesian optimization is now popular for hyperparameter optimization, since it reduces the number of validation error evaluations required. Suppose that we are given a collection of datasets on which hyperparameters are already tuned by either humans with ...

متن کامل

Gradient-based Hyperparameter Optimization through Reversible Learning

Tuning hyperparameters of learning algorithms is hard because gradients are usually unavailable. We compute exact gradients of cross-validation performance with respect to all hyperparameters by chaining derivatives backwards through the entire training procedure. These gradients allow us to optimize thousands of hyperparameters, including step-size and momentum schedules, weight initialization...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2023

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v37i7.26070